# Sparse Attention

Lsg Legal Small Uncased 4096
A compact version of LEGAL-BERT, employing Local+Sparse+Global attention mechanism (LSG) for efficient long-sequence processing
Large Language Model Transformers English
L
ccdv
1,088
0
Bigbird Base Trivia Itc
Apache-2.0
A fine-tuned model based on bigbird-roberta-base, optimized for trivia QA tasks with long sequence processing support.
Question Answering System English
B
google
1,049
8
Bigbird Pegasus Large Bigpatent
Apache-2.0
BigBird is a Transformer model based on sparse attention, capable of processing sequences up to 4096 in length, suitable for tasks like long document summarization.
Text Generation Transformers English
B
google
945
40
Bigbird Roberta Base Finetuned App
MIT
A Transformer model based on sparse attention, fine-tuned specifically for mobile app description classification tasks
Text Classification Transformers English
B
nsi319
15
0
Bigbird Pegasus Large Arxiv
Apache-2.0
BigBird is a Transformer model based on sparse attention, capable of handling longer sequences, suitable for tasks like long document summarization.
Text Generation Transformers English
B
google
8,528
61
Bigbird Pegasus Large Pubmed
Apache-2.0
BigBirdPegasus is a Transformer model based on sparse attention, capable of handling longer sequences, especially suitable for long document summarization tasks.
Text Generation Transformers English
B
google
2,031
47
Bigbird Roberta Large
Apache-2.0
BigBird is a Transformer model based on sparse attention, capable of processing sequences up to 4096 tokens long, suitable for long document tasks.
Large Language Model English
B
google
1,152
27
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase